Gradient - Descent / ClusteringArchitecture forFinite State Machine Induction
نویسندگان
چکیده
Although recurrent neural nets have been moderately successful in learning to emulate nite-state machines (FSMs), the continuous internal state dynamics of a neural net are not well matched to the discrete behavior of an FSM. We describe an architecture, called DOLCE, that allows discrete states to evolve in a net as learning progresses. dolce consists of a standard recurrent neural net trained by gradient descent and an adaptive clustering technique that quantizes the state space. dolce is based on the assumption that a nite set of discrete internal states is required for the task, and that the actual network state belongs to this set but has been corrupted by noise due to inaccuracy in the weights. dolce learns to recover the discrete state with maximum a posteriori probability from the noisy state. Simulations show that dolce leads to a signiicant improvement in generalization performance over earlier neural net approaches to FSM induction.
منابع مشابه
A Unified Gradient-Descent/Clustering Architecture for Finite State Machine Induction
Although recurrent neural nets have been moderately successful in learning to emulate finite-state machines (FSMs), the continuous internal state dynamics of a neural net are not well matched to the discrete behavior of an FSM. We describe an architecture, called DOLCE, that allows discrete states to evolve in a net as learning progresses. DOLCE consists of a standard recurrent neural net train...
متن کاملLess Regret via Online Conditioning
We analyze and evaluate an online gradient descent algorithm with adaptive per-coordinate adjustment of learning rates. Our algorithm can be thought of as an online version of batch gradient descent with a diagonal preconditioner. This approach leads to regret bounds that are stronger than those of standard online gradient descent for general online convex optimization problems. Experimentally,...
متن کاملOn-the-Fly Learning in a Perpetual Learning Machine
Despite the promise of brain-inspired machine learning, deep neural networks (DNN) have frustratingly failed to bridge the deceptively large gap between learning and memory. Here, we introduce a Perpetual Learning Machine; a new type of DNN that is capable of brain-like dynamic ‘on the fly’ learning because it exists in a self-supervised state of Perpetual Stochastic Gradient Descent. Thus, we ...
متن کاملMachine Learning Applications in Fantasy Basketball
This paper is an attempt to apply machine learning to fantasy sports in order to gain an edge over the average player. Basketball players’ fantasy scores were predicted using a linear regression algorithm and stochastic gradient descent as well as a naive bayes classifier with discretized state space. A team of eight players was then selected by framing the problem as a constraint satisfaction ...
متن کاملUnifying the Stochastic Spectral Descent for Restricted Boltzmann Machines with Bernoulli or Gaussian Inputs
Stochastic gradient descent based algorithms are typically used as the general optimization tools for most deep learning models. A Restricted Boltzmann Machine (RBM) is a probabilistic generative model that can be stacked to construct deep architectures. For RBM with Bernoulli inputs, non-Euclidean algorithm such as stochastic spectral descent (SSD) has been specifically designed to speed up th...
متن کامل